# Text feature extraction

Amber Base
Apache-2.0
Amber Base is a Japanese-English sentence encoder model based on modernbert-ja-130m, specializing in sentence similarity calculation and feature extraction tasks.
Text Embedding Supports Multiple Languages
A
retrieva-jp
213
3
Context Skill Extraction Base
This is a model trained based on sentence-transformers, capable of mapping sentences and paragraphs into a 768-dimensional dense vector space, suitable for various tasks such as semantic text similarity calculation and semantic search.
Text Embedding
C
TechWolf
189
5
Snowflake Arctic Embed M V1.5
Apache-2.0
Snowflake Arctic Embed M v1.5 is an efficient sentence embedding model, focusing on sentence similarity calculation and feature extraction tasks.
Text Embedding
S
Snowflake
219.46k
58
Bge Micro V2
bge_micro is a sentence embedding model based on sentence-transformers, focusing on sentence similarity calculation and feature extraction tasks.
Text Embedding Transformers
B
SmartComponents
468
2
St Polish Kartonberta Base Alpha V1
This is a Polish sentence transformer model based on the KartonBERTa architecture, primarily used for sentence similarity calculation and feature extraction tasks.
Text Embedding Transformers Other
S
OrlikB
3,494
3
Bge Micro V2
MIT
bge_micro is a lightweight model focused on sentence similarity calculation, suitable for various natural language processing tasks.
Text Embedding Transformers
B
TaylorAI
248.53k
46
Bge Micro
bge_micro is a lightweight sentence similarity calculation model based on transformer architecture, specifically designed for efficient feature extraction and sentence similarity tasks.
Text Embedding Transformers
B
TaylorAI
1,799
23
Unsup Simcse Ja Base
This is an unsupervised SimCSE-based Japanese sentence embedding model, specifically designed for generating high-quality Japanese sentence embeddings.
Text Embedding Transformers Japanese
U
cl-nagoya
190
2
Distilbert Base Uncased Becas 2
Apache-2.0
Model fine-tuned on distilbert-base-uncased using the becasv2 dataset, with a validation loss of 5.9506
Large Language Model Transformers
D
Evelyn18
16
0
Large
Apache-2.0
A Transformer model pre-trained on English corpus using ELECTRA-like objective functions, learning intrinsic representations of English language through self-supervised methods.
Large Language Model Transformers English
L
funnel-transformer
190
2
Finetune Data Skills
Apache-2.0
A BERT-based fine-tuned model for data skills, suitable for specific NLP tasks
Large Language Model Transformers
F
dpasch01
26
0
Xlarge
Apache-2.0
Funnel Transformer is an English text pre-training model based on self-supervised learning, adopting objectives similar to ELECTRA, achieving efficient language processing by filtering sequence redundancy.
Large Language Model Transformers English
X
funnel-transformer
31
1
Intermediate
Apache-2.0
Transformer model pre-trained on English corpus using ELECTRA-like objectives, acquiring text representations through self-supervised learning
Large Language Model Transformers English
I
funnel-transformer
24
0
Albert Base Chinese Cluecorpussmall
A Chinese ALBERT model pre-trained on CLUECorpusSmall, trained by the UER-py framework, suitable for Chinese text processing tasks.
Large Language Model Transformers Chinese
A
uer
7,203
37
Small
Apache-2.0
Transformer model pre-trained on English corpus using ELECTRA-like objectives, suitable for text feature extraction and downstream task fine-tuning
Large Language Model Transformers English
S
funnel-transformer
6,084
5
Gpt2 Large Bne
Apache-2.0
GPT2-large Spanish language model trained on data from the National Library of Spain (BNE)
Large Language Model Transformers Spanish
G
PlanTL-GOB-ES
899
12
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase